Manipulating and Measuring Model Interpretability

نویسندگان

  • Forough Poursabzi-Sangdeh
  • Daniel G. Goldstein
  • Jake M. Hofman
  • Jennifer Wortman Vaughan
  • Hanna M. Wallach
چکیده

Despite a growing body of research focused on creating interpretable machine learning methods, there have been few empirical studies verifying whether interpretable methods achieve their intended effects on end users. We present a framework for assessing the effects of model interpretability on users via pre-registered experiments in which participants are shown functionally identical models that vary in factors thought to influence interpretability. Using this framework, we ran a sequence of large-scale randomized experiments, varying two putative drivers of interpretability: the number of features and the model transparency (clear or black-box). We measured how these factors impact trust in model predictions, the ability to simulate a model, and the ability to detect a model’s mistakes. We found that participants who were shown a clear model with a small number of features were better able to simulate the model’s predictions. However, we found no difference in multiple measures of trust and found that clear models did not improve the ability to correct mistakes. These findings suggest that interpretability research could benefit from more emphasis on empirically verifying that interpretable models achieve all their intended effects.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Evaluating Regularized Anchor Words

We perform a comprehensive examination of the recently proposed anchor method for topic model inference using topic interpretability and held-out likelihood measures. After measuring the sensitivity to the anchor selection process, we incorporate L2 and Beta regularization into the optimization objective in the recovery step. Preliminary results show that L2 improves heldout likelihood, and Bet...

متن کامل

SECURING INTERPRETABILITY OF FUZZY MODELS FOR MODELING NONLINEAR MIMO SYSTEMS USING A HYBRID OF EVOLUTIONARY ALGORITHMS

In this study, a Multi-Objective Genetic Algorithm (MOGA) is utilized to extract interpretable and compact fuzzy rule bases for modeling nonlinear Multi-input Multi-output (MIMO) systems. In the process of non- linear system identi cation, structure selection, parameter estimation, model performance and model validation are important objectives. Furthermore, se- curing low-level and high-level ...

متن کامل

Discussions on Interpretability of Fuzzy Systems using Simple Examples

Two conflicting goals are often involved in the design of fuzzy rule-based systems: Accuracy maximization and interpretability maximization. A number of approaches have been proposed for finding a fuzzy rule-based system with a good accuracy-interpretability tradeoff. Formulation of the accuracy maximization is usually straightforward in each application area of fuzzy rule-based systems such as...

متن کامل

Measuring interpretability in rule-based classification systems

The ‘hnique selling point” of fuzzy systems is usually the interpretability of its rule base. However, very often only the U C C U T U C ~ of the rule base is measured and used to compare a fuzzy system to other solutions. We have suggested an index to measurz the interpretability of fuzzy rule bases for classification problems. However, the index can be used to describe the interpretability of...

متن کامل

Improving the interpretability of data-driven evolving fuzzy systems

This paper develops methods for reducing the complexity and, thereby, improving the linguistic interpretability of Takagi-Sugeno fuzzy systems that are learned online in a data-driven, incremental way. In order to ensure the transparency of the evolving fuzzy system at any time, complexity reduction must be performed in an online mode as well. Our methods are evaluated on high-dimensional data ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1802.07810  شماره 

صفحات  -

تاریخ انتشار 2018